Goto

Collaborating Authors

 real-time transcription


Dialogue in Resonance: An Interactive Music Piece for Piano and Real-Time Automatic Transcription System

Bang, Hayeon, Kwon, Taegyun, Nam, Juhan

arXiv.org Artificial Intelligence

This paper presents , an interactive music piece for a human pianist and a computer-controlled piano that integrates real-time automatic music transcription into a score-driven framework. Unlike previous approaches that primarily focus on improvisation-based interactions, our work establishes a balanced framework that combines composed structure with dynamic interaction. Through real-time automatic transcription as its core mechanism, the computer interprets and responds to the human performer's input in real time, creating a musical dialogue that balances compositional intent with live interaction while incorporating elements of unpredictability. In this paper, we present the development process from composition to premiere performance, including technical implementation, rehearsal process, and performance considerations.


Evaluation of real-time transcriptions using end-to-end ASR models

Arriaga, Carlos, Pozo, Alejandro, Conde, Javier, Alonso, Alvaro

arXiv.org Artificial Intelligence

Automatic Speech Recognition (ASR) or Speech-to-text (STT) has greatly evolved in the last few years. Traditional architectures based on pipelines have been replaced by joint end-to-end (E2E) architectures that simplify and streamline the model training process. In addition, new AI training methods, such as weak-supervised learning have reduced the need for high-quality audio datasets for model training. However, despite all these advancements, little to no research has been done on real-time transcription. In real-time scenarios, the audio is not pre-recorded, and the input audio must be fragmented to be processed by the ASR systems. To achieve real-time requirements, these fragments must be as short as possible to reduce latency. However, audio cannot be split at any point as dividing an utterance into two separate fragments will generate an incorrect transcription. Also, shorter fragments provide less context for the ASR model. For this reason, it is necessary to design and test different splitting algorithms to optimize the quality and delay of the resulting transcription. In this paper, three audio splitting algorithms are evaluated with different ASR models to determine their impact on both the quality of the transcription and the end-to-end delay. The algorithms are fragmentation at fixed intervals, voice activity detection (VAD), and fragmentation with feedback. The results are compared to the performance of the same model, without audio fragmentation, to determine the effects of this division. The results show that VAD fragmentation provides the best quality with the highest delay, whereas fragmentation at fixed intervals provides the lowest quality and the lowest delay. The newly proposed feedback algorithm exchanges a 2-4% increase in WER for a reduction of 1.5-2s delay, respectively, to the VAD splitting.


How To Transcribe Streams of Audio Data in Real-Time with Python?

#artificialintelligence

In my previous blog posts, I went through the AssemblyAI speech-to-text API. I tried its core transcription service and played with some of its cool AI-powered features: the content moderation feature that spots sensitive topics and the topic detection feature that extracts the subjects that are spoken about in each audio segment. You can check it out here. The code is also available on Github.) These experiments are all performed offline and take some time to run in order to generate the output.


Online Sequence Alignment for Real-Time Audio Transcription by Non-Experts

Lasecki, Walter S. (University of Rochester) | Miller, Christopher D. (University of Rochester) | Borrello, Donato (Univeristy of Rochester) | Bigham, Jeffrey P. (University of Rochester)

AAAI Conferences

Real-time transcription provides deaf and hard of hearing people visual access to spoken content, such as classroom instruction, and other live events. Currently, the only reliable source of real-time transcriptions are expensive, highly-trained experts who are able to keep up with speaking rates. Automatic speech recognition is cheaper but produces too many errors in realistic settings. We introduce a new approach in which partial captions from multiple non-experts are combined to produce a high-quality transcription in real-time. We demonstrate the potential of this approach with data collected from 20 non-expert captionists.